74 research outputs found
AutoQS v1: automatic parametrization of QuickSampling based on training images analysis
Multiple-point geostatistics are widely used to simulate
complex spatial structures based on a training image. The practical
applicability of these methods relies on the possibility of finding optimal
training images and parametrization of the simulation algorithms. While
methods for automatically selecting training images are available,
parametrization can be cumbersome. Here, we propose to find an optimal set
of parameters using only the training image as input. The difference between
this and previous work that used parametrization optimization is that it
does not require the definition of an objective function. Our approach is
based on the analysis of the errors that occur when filling artificially
constructed patterns that have been borrowed from the training image. Its
main advantage is to eliminate the risk of overfitting an objective
function, which may result in variance underestimation or in verbatim copy
of the training image. Since it is not based on optimization, our approach
finds a set of acceptable parameters in a predictable manner by using the
knowledge and understanding of how the simulation algorithms work. The
technique is explored in the context of the recently developed QuickSampling
algorithm, but it can be easily adapted to other pixel-based multiple-point
statistics algorithms using pattern matching, such as direct sampling or
single normal equation simulation (SNESIM).</p
Устройство для перемещения датчиков в магнитном поле малогабаритного бетатрона
Рассматривается возможность увеличения точности измерений характеристик магнитного поля посредством более точной установки датчиков в исследуемой точке
GeantV: Results from the prototype of concurrent vector particle transport simulation in HEP
Full detector simulation was among the largest CPU consumer in all CERN
experiment software stacks for the first two runs of the Large Hadron Collider
(LHC). In the early 2010's, the projections were that simulation demands would
scale linearly with luminosity increase, compensated only partially by an
increase of computing resources. The extension of fast simulation approaches to
more use cases, covering a larger fraction of the simulation budget, is only
part of the solution due to intrinsic precision limitations. The remainder
corresponds to speeding-up the simulation software by several factors, which is
out of reach using simple optimizations on the current code base. In this
context, the GeantV R&D project was launched, aiming to redesign the legacy
particle transport codes in order to make them benefit from fine-grained
parallelism features such as vectorization, but also from increased code and
data locality. This paper presents extensively the results and achievements of
this R&D, as well as the conclusions and lessons learnt from the beta
prototype.Comment: 34 pages, 26 figures, 24 table
A high-resolution image time series of the Gorner Glacier – Swiss Alps – derived from repeated unmanned aerial vehicle surveys
Modern drone technology provides an efficient means to monitor the response
of alpine glaciers to climate warming. Here we present a new topographic
dataset based on images collected during 10 UAV surveys of the
Gorner Glacier glacial system
(Switzerland) carried out approximately every 2 weeks throughout the summer
of 2017. The final products, available at https://doi.org/10.5281/zenodo.2630456
(Benoit et al., 2018), consist of a series of 10 cm resolution orthoimages,
digital elevation models of the glacier surface, and maps of ice surface displacement. Used on its own, this dataset allows mapping of the glacier
and monitoring surface velocities over the summer at a very high spatial
resolution. Coupled with a classification or feature detection algorithm, it
enables the extraction of structures such as surface drainage networks,
debris, or snow cover. The approach we present can be used in the future to
gain insights into ice flow dynamics.</p
A routing and resource preservation strategy for traffic engineering in communication networks
International audience; This paper presents a method for dynamic load balancing in data networks. When multiple routes are available, it determines their load shares as a function of a composite metric that takes into account the paths? length and load. A general resource preservation mechanism is also presented that complements the proposed random routing strategy when the network is heavily loaded. We compare our approach with trunk reservation in the particular case of fully meshed networks and evaluate its performance in any network, where an equivalent mechanism is missing. We validate our approach by means of simulation and provide insights on the routing solutions that it obtains.
Document type: Part of book or chapter of boo
Impact of peer-to-peer traffic on the efficiency of optical packet rings
International audienc
- …